Unsupervised Video Object Segmentation via Weak User Interaction and Temporal Modulation

نویسندگان

چکیده

In unsupervised video object segmentation (UVOS), the whole might segment wrong target due to lack of initial prior information. Also, in semi-supervised (SVOS), frame with a fine-grained pixel-level mask is essential good accuracy. It expensive and laborious provide accurate masks for each training sequence. To address this issue, We present weak user interactive UVOS approach guided by simple human-made rectangle annotation frame. first interactively draw region interest rectangle, then we leverage RCNN (region-based convolutional neural networks) method generate set coarse reference labels subsequent propagations. establish temporal correspondence between coherent frames, further design two novel modulation modules enhance representations. compute earth mover's distance (EMD)-based similarity frames mine co-occurrent objects images, which used modulate representation highlight foreground target. cross-squeeze module emphasize features across helps representation. augment temporally modulated representations original obtain compositive spatio-temporal information, producing more (VOS) model. The experimental results on both SVOS datasets including Davis2016, FBMS, Youtube-VOS, Davis2017, show that our yields favorable accuracy complexity. related code available.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Efficient Video Object Segmentation via Network Modulation

Video object segmentation targets at segmenting a specific object throughout a video sequence, given only an annotated first frame. Recent deep learning based approaches find it effective by fine-tuning a general-purpose segmentation model on the annotated frame using hundreds of iterations of gradient descent. Despite the high accuracy these methods achieve, the fine-tuning process is ineffici...

متن کامل

Temporal Video Segmentation Using Unsupervised

This paper proposes a content-based temporal video segmentation system that integrates syntactic (domain-independent) and semantic (domain-dependent) features for automatic management of video data. Temporal video segmentation includes scene change detection and shot classiication. The proposed scene change detection method consists of two steps: detection and tracking of semantic objects of in...

متن کامل

Video Object Segmentation Without Temporal Information

Video Object Segmentation, and video processing in general, has been historically dominated by methods that rely on the temporal consistency and redundancy in consecutive video frames. When the temporal smoothness is suddenly broken, such as when an object is occluded, or some frames are missing in a sequence; the result of these methods can deteriorate significantly or they may not even produc...

متن کامل

A Neural Network based Scheme for Unsupervised Video Object Segmentation

In this paper, we proposed a neural network based scheme for performing unsupervised video object segmentation, especially for videophone or videoconferencing applications. The procedure includes (a) a training algorithm for adapting the network weights to the current condition, (b) a Maximum A Posteriori (MAP) estimation procedure for optimally selecting the most representative data of the cur...

متن کامل

Unsupervised Semantic Object Segmentation of Stereoscopic Video Sequences

In this paper, we present an efficient technique for unsupervised semantically meaningful object segmentation of stereoscopic video sequences. By this technique we achieve to extract semantic objects using the additional information a stereoscopic pair of frames provides. Each pair is analyzed and the disparity field, occluded areas and depth map are estimated. The key algorithm, which is appli...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Chinese Journal of Electronics

سال: 2023

ISSN: ['1022-4653', '2075-5597']

DOI: https://doi.org/10.23919/cje.2022.00.139